Geometry-aware principal component analysis for symmetric positive definite matrices

نویسندگان
چکیده

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Symmetric Positive-Definite Matrices: From Geometry to Applications and Visualization

In many engineering applications that use tensor analysis, such as tensor imaging, the underlying tensors have the characteristic of being positive definite. It might therefore be more appropriate to use techniques specially adapted to such tensors. We will describe the geometry and calculus on the Riemannian symmetric space of positive-definite tensors. First, we will explain why the geometry,...

متن کامل

DDtBe for Band Symmetric Positive Definite Matrices

We present a new parallel factorization for band symmetric positive definite (s.p.d) matrices and show some of its applications. Let A be a band s.p.d matrix of order n and half bandwidth m. We show how to factor A as A =DDt Be using approximately 4nm2 jp parallel operations where p =21: is the number of processors. Having this factorization, we improve the time to solve Ax = b by a factor of m...

متن کامل

Riemannian geometry on positive definite matrices

The Riemannian metric on the manifold of positive definite matrices is defined by a kernel function φ in the form K D(H,K) = ∑ i,j φ(λi, λj) −1TrPiHPjK when ∑ i λiPi is the spectral decomposition of the foot point D and the Hermitian matrices H,K are tangent vectors. For such kernel metrics the tangent space has an orthogonal decomposition. The pull-back of a kernel metric under a mapping D 7→ ...

متن کامل

Wasserstein Riemannian Geometry of Positive-definite Matrices∗

The Wasserstein distance on multivariate non-degenerate Gaussian densities is a Riemannian distance. After reviewing the properties of the distance and the metric geodesic, we derive an explicit form of the Riemannian metrics on positive-definite matrices and compute its tensor form with respect to the trace scalar product. The tensor is a matrix, which is the solution of a Lyapunov equation. W...

متن کامل

Supervised LogEuclidean Metric Learning for Symmetric Positive Definite Matrices

Metric learning has been shown to be highly effective to improve the performance of nearest neighbor classification. In this paper, we address the problem of metric learning for symmetric positive definite (SPD) matrices such as covariance matrices, which arise in many real-world applications. Naively using standard Mahalanobis metric learning methods under the Euclidean geometry for SPD matric...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Machine Learning

سال: 2016

ISSN: 0885-6125,1573-0565

DOI: 10.1007/s10994-016-5605-5